Efficient Estimation of the Value of Information in Monte Carlo Models
نویسندگان
چکیده
The expected value of information (EVI) is the most powerful measure of sensitivity to uncer tainty in a decision model: it measures the potential of information to improve the decision, and hence measures the expected value of the outcome. Stan dard methods for computing EVI use discrete vari ables and are computationally intractable for models that contain more than a few variables. Monte Carlo simulation provides the basis for more tractable evaluation of large predictive models with continu ous and discrete variables, but so far computation of EVI in a Monte Carlo setting also has appeared impractical. We introduce an approximate approach based on preposterior analysis for estimating EVI in Monte Carlo models. Our method uses a linear approximation to the value function and multiple linear regression to estimate the linear model from the samples. The approach is efficient and practical for extremely large models. It allows easy estima tion of EVI for perfect or panial information on individual variables or on combinations of variables. We illustrate its implementation within Demos (a decision modeling system), and its application to a large model for crisis transponation planning. 1.0 EVI: What's so, and What's New A ny model is inevitably a simplification of reality, and most of its input quantities are invariably uncertain. Sensi tivity analysis identifies which sources of uncertainty in a model affect its outputs most significantly. In this way, it helps a decision maker focus attention on what assump tions really matter. It also helps a decision modeler to assign priorities to his efforts to improve, refine, or extend his model by identifying those variables for which it will be most valuable to find more complete data, to interview more knowledgeable experts, or to build more elaborate submodels. The expected value of information (EVI) on a variable xi measures the expected increase in value y if we learn new information about xi and make a decision with higher expected value in light of that information. It is the most powerful method of sensitivity analysis because it ana lyzes a variable's importance in terms of the overall pre scription for action, and it expresses that importance in the utility or value units of the problem. Other methods, such as rank-order correlation, express importance in terms of the correlation between an uncertain variable and the out put of the decision model. There are many cases where a variable can show high sensitivity in this way, yet still have no effect on the selection of an optimal decision. Deterministic perturbation measures importance in utility or value units, but it ignores nonlinearities and interactions among variables, and also fails to measure a variable's importance in terms of that variable's ability to change the recommended decision. 120 Chavez and Henrion One calculates EVPI (Expected Value of Perfect Infor· mation) in discrete models by rolling back the decision tree. The computation itself is straightforward in the sense that, to compute EVI, one simply places at the front of the tree the chance variables to be observed. The EVPI is computed as the difference between the expected value computed for this scenario and the expected value for the regular tree, without observations. Computing EVI with continuous variables is less intuitive, because we have no tidy way of reversing the uncertainty, unlike the discrete case. Yet continuous models are increasingly the norm for risk and decision analysis, first because discretizing inherently continuous variables intro duces unnecessary approximation, and second because Monte Carlo methods and their variants (e.g., Latin hyper cube) generate tractable, highly efficient solutions to pre dictive models that contain thousands of variables. An especially useful feature of the Monte Carlo method is that, for a specified error, the computational complexity increases linearly in the number of uncertain variables [Morgan and Henrion, 1990]. Exact methods require com putation time that is exponential in the number of vari ables. There is thus a need to develop flexible, efficient methods for computing EVI on continuous variables in a Monte Carlo setting. A ftexible method has (I) the ability to com pute EVI on single variables or on any combination of variables, and (2) the ability to compute both perfect and partial values of information. Perfect information removes uncertainty entirely. Partial information reduces uncertainty. We present a general framework for calculating EVI based on preposterior analysis. Using that framework, we develop a technique for computing EVI that depends on a linear approximation to the value function and on multiple linear regression to estimate the constants for the linear function. We also discuss a heuristic method for measuring the value of partial information in terms of what we call the relative information multiple (RIM). We have implemented these methods in detachable computational modules using Demos, a decision modeling system from Lumina, Inc., Palo Alto, CA. We demonstrate their use on a large model to aid in military transportation crisis plan ning. 2.0 Framework A decision model consists of a set of n state variables x1, •• ,xn, which we will denote by X. The decision maker has control of a decision variable D, which can assume one of m possible values d1 , .. ,dm. The value or utility func tion v(X,di) expresses the payoff to the decision maker when X obtains and decision di is chosen. In a typical decision model, the state variables are uncer tain. We express prior knowledge about X in the form of a probability distribution, denoted {XI �},where �denotes a prior state of information. The optimal Bayes' decision maximizing the expected value1 is given by l = Arg max ( X d I l' ) d v ( ' ) .., . The optimal decision given perfect information on state variable x , denoted d* _., is d• x __ Arg max d (v(X,d)Jx,�) We define EVPI on x as EVPI (x) = (v (X, l x) I �){v (X, d*) I �). In a similar fashion, we define the optimal expected value decision given the revelation of evidence e, d* e• as le= Arg d max (v(X,d)Je.�) Then the EVI for evidence e is EVI(e) = (v(X,d.e)l�)-(v(X,l)l�) 2.1 Binary decisions and Function z Let us consider a simplified decision problem with two decision alternatives: one of them is the optimal Bayes' decision d*; the other we denote tr". In view of the uncertainty in the state variables, there must exist uncertainty in the outputs as well. Thus, for each 1. We use Howard's inferential notation (see, for example, Howard, 1970). {XIS) denotes the probability density of X condi tional on S; (XIS) denotes the expectation of X conditional on S. Efficient Estimation of the Value of Information in Monte Carlo Models 121 decision d;, there exists a unique probability distribution on value { v(X,d;)l �} (see Figure 1). For notational conve nience we let v(d) = (v(X, d) I�). We now define z = v(X,l) -v(X,tf). Function z is the pivotal element in our framework for computing EVI because it describes the difference in value between the best and second-best decisions. In Figure 2, we have graphed the probability distribution of z. The shaded area represents the total probability of making a bad decision, i.e., doing d' when cY would yield higher value. Exploiting information encoded in the shaded, neg ative portion of the z distribution's curve will provide the necessary clues to compute EVPI and EVI. FIGURE 1. Probability �istributions on value for the two decisions d and cr-. Pro density {v(X,tf)[�} {v(X,l)[s} FIGURE 2. Function z: the difference in value between the best and second-best decisions.
منابع مشابه
Positive-Shrinkage and Pretest Estimation in Multiple Regression: A Monte Carlo Study with Applications
Consider a problem of predicting a response variable using a set of covariates in a linear regression model. If it is a priori known or suspected that a subset of the covariates do not significantly contribute to the overall fit of the model, a restricted model that excludes these covariates, may be sufficient. If, on the other hand, the subset provides useful information, shrinkage meth...
متن کاملA Monte Carlo-Based Search Strategy for Dimensionality Reduction in Performance Tuning Parameters
Redundant and irrelevant features in high dimensional data increase the complexity in underlying mathematical models. It is necessary to conduct pre-processing steps that search for the most relevant features in order to reduce the dimensionality of the data. This study made use of a meta-heuristic search approach which uses lightweight random simulations to balance between the exploitation of ...
متن کاملApplication of Monte Carlo Simulation in the Assessment of European Call Options
In this paper, the pricing of a European call option on the underlying asset is performed by using a Monte Carlo method, one of the powerful simulation methods, where the price development of the asset is simulated and value of the claim is computed in terms of an expected value. The proposed approach, applied in Monte Carlo simulation, is based on the Black-Scholes equation which generally def...
متن کاملApplying Point Estimation and Monte Carlo Simulation Methods in Solving Probabilistic Optimal Power Flow Considering Renewable Energy Uncertainties
The increasing penetration of renewable energy results in changing the traditional power system planning and operation tools. As the generated power by the renewable energy resources are probabilistically changed, the certain power system analysis tolls cannot be applied in this case. Probabilistic optimal power flow is one of the most useful tools regarding the power system analysis in presen...
متن کاملRanking Bank Branches with Interval Data By IAHP and TOPSIS
This paper proposes a method for ranking decision making units (DMUs) using some of the multiple criteria decision making / multiple attribute decision making (MCDM /MADM) techniques, namely, interval analytic hierarchy process (IAHP) and the technique for order preference by similarity to an ideal solution (TOPSIS). Since the efficiency score of unity is assigned to ...
متن کاملEstimation for the Type-II Extreme Value Distribution Based on Progressive Type-II Censoring
In this paper, we discuss the statistical inference on the unknown parameters and reliability function of type-II extreme value (EVII) distribution when the observed data are progressively type-II censored. By applying EM algorithm, we obtain maximum likelihood estimates (MLEs). We also suggest approximate maximum likelihood estimators (AMLEs), which have explicit expressions. We provide Bayes ...
متن کامل